Proxmox Cluster Basics

Overview of how Proxmox VE clusters work, including quorum, networking, and operational constraints

created: Sat Mar 14 2026 00:00:00 GMT+0000 (Coordinated Universal Time) updated: Sat Mar 14 2026 00:00:00 GMT+0000 (Coordinated Universal Time) #proxmox#virtualization#clustering

Introduction

A Proxmox VE cluster groups multiple Proxmox nodes into a shared management domain. This allows centralized administration of virtual machines, containers, storage definitions, and optional high-availability workflows.

Purpose

Use a Proxmox cluster when you want:

  • Centralized management for multiple hypervisor nodes
  • Shared visibility of guests, storage, and permissions
  • Live migration or controlled workload movement between nodes
  • A foundation for HA services backed by shared or replicated storage

Architecture Overview

A Proxmox cluster relies on several core components:

  • pvecm: the cluster management tool used to create and join clusters
  • Corosync: provides the cluster communication layer
  • pmxcfs: the Proxmox cluster file system used to distribute cluster configuration
  • Quorum: majority voting used to protect cluster consistency

Important operational behavior:

  • Each node normally has one vote
  • A majority of votes must be online for state-changing operations
  • Loss of quorum causes the cluster to become read-only for protected operations

Cluster Design Notes

Network requirements

Proxmox expects a reliable low-latency network for cluster traffic. Corosync is sensitive to packet loss, jitter, and unstable links. In homelabs, this generally means wired LAN links, stable switching, and avoiding Wi-Fi for cluster communication.

Odd node counts

Three nodes is the common minimum for a healthy quorum-based design. Two-node designs can work, but they need extra planning such as a QDevice or acceptance of reduced fault tolerance.

Storage considerations

Clustering does not automatically provide shared storage. Features such as live migration and HA depend on storage design:

  • Shared storage: NFS, iSCSI, Ceph, or other shared backends
  • Replicated local storage: possible for some workflows, but requires careful planning
  • Backup storage: separate from guest runtime storage

Configuration Example

Create a new cluster on the first node:

pvecm create lab-cluster

Check cluster status:

pvecm status

Join another node to the cluster from that node:

pvecm add 192.0.2.10

Use placeholder management addresses in documentation and never expose real administrative IPs publicly.

Troubleshooting Tips

Cluster is read-only

  • Check quorum status with pvecm status
  • Look for network instability between nodes
  • Verify time synchronization and general host health

Node join fails

  • Confirm name resolution and basic IP reachability
  • Make sure cluster traffic is not filtered by a firewall
  • Verify the node is not already part of another cluster

Random cluster instability

  • Review packet loss, duplex mismatches, and switch reliability
  • Keep corosync on stable wired links with low latency
  • Separate heavy storage replication traffic from cluster messaging when possible

Best Practices

  • Use at least three voting members for a stable quorum model
  • Keep cluster traffic on reliable wired networking
  • Document node roles, storage backends, and migration dependencies
  • Treat the Proxmox management network as a high-trust segment
  • Test backup and restore separately from cluster failover assumptions

References